7 research outputs found
The SEN1-2 Dataset for Deep Learning in SAR-Optical Data Fusion
While deep learning techniques have an increasing impact on many technical
fields, gathering sufficient amounts of training data is a challenging problem
in remote sensing. In particular, this holds for applications involving data
from multiple sensors with heterogeneous characteristics. One example for that
is the fusion of synthetic aperture radar (SAR) data and optical imagery. With
this paper, we publish the SEN1-2 dataset to foster deep learning research in
SAR-optical data fusion. SEN1-2 comprises 282,384 pairs of corresponding image
patches, collected from across the globe and throughout all meteorological
seasons. Besides a detailed description of the dataset, we show exemplary
results for several possible applications, such as SAR image colorization,
SAR-optical image matching, and creation of artificial optical images from SAR
input data. Since SEN1-2 is the first large open dataset of this kind, we
believe it will support further developments in the field of deep learning for
remote sensing as well as multi-sensor data fusion.Comment: accepted for publication in the ISPRS Annals of the Photogrammetry,
Remote Sensing and Spatial Information Sciences (online from October 2018
SEN12MS -- A Curated Dataset of Georeferenced Multi-Spectral Sentinel-1/2 Imagery for Deep Learning and Data Fusion
The availability of curated large-scale training data is a crucial factor for
the development of well-generalizing deep learning methods for the extraction
of geoinformation from multi-sensor remote sensing imagery. While quite some
datasets have already been published by the community, most of them suffer from
rather strong limitations, e.g. regarding spatial coverage, diversity or simply
number of available samples. Exploiting the freely available data acquired by
the Sentinel satellites of the Copernicus program implemented by the European
Space Agency, as well as the cloud computing facilities of Google Earth Engine,
we provide a dataset consisting of 180,662 triplets of dual-pol synthetic
aperture radar (SAR) image patches, multi-spectral Sentinel-2 image patches,
and MODIS land cover maps. With all patches being fully georeferenced at a 10 m
ground sampling distance and covering all inhabited continents during all
meteorological seasons, we expect the dataset to support the community in
developing sophisticated deep learning-based approaches for common tasks such
as scene classification or semantic segmentation for land cover mapping.Comment: accepted for publication in the ISPRS Annals of the Photogrammetry,
Remote Sensing and Spatial Information Sciences (online from September 2019
Enhancing mobile camera pose estimation through the inclusion of sensors
Thesis (MSc)--Stellenbosch University, 2014.ENGLISH ABSTRACT: Monocular structure from motion (SfM) is a widely researched problem, however
many of the existing approaches prove to be too computationally expensive for use
on mobile devices. In this thesis we investigate how inertial sensors can be used
to increase the performance of SfM algorithms on mobile devices.
Making use of the low cost inertial sensors found on most mobile devices we
design and implement an extended Kalman filter (EKF) to exploit their complementary
nature, in order to produce an accurate estimate of the attitude of the
device. We make use of a quaternion based system model in order to linearise the
measurement stage of the EKF, thus reducing its computational complexity. We
use this attitude estimate to enhance the feature tracking and camera localisation
stages in our SfM pipeline.
In order to perform feature tracking we implement a hybrid tracking algorithm
which makes use of Harris corners and an approximate nearest neighbour search to
reduce the search space for possible correspondences. We increase the robustness
of this approach by using inertial information to compensate for inter-frame camera
rotation. We further develop an efficient bundle adjustment algorithm which
only optimises the pose of the previous three key frames and the 3D map points
common between at least two of these frames. We implement an optimisation
based localisation algorithm which makes use of our EKF attitude estimate and
the tracked features, in order to estimate the pose of the device relative to the 3D
map points. This optimisation is performed in two steps, the first of which optimises
only the translation and the second optimises the full pose. We integrate the
aforementioned three sub-systems into an inertial assisted pose estimation pipeline.
We evaluate our algorithms with the use of datasets captured on the iPhone
5 in the presence of a Vicon motion capture system for ground truth data. We
find that our EKF can estimate the device’s attitude with an average dynamic
accuracy of ±5°. Furthermore, we find that the inclusion of sensors into the visual
pose estimation pipeline can lead to improvements in terms of robustness and
computational efficiency of the algorithms and are unlikely to negatively affect the
accuracy of such a system. Even though we managed to reduce execution time
dramatically, compared to typical existing techniques, our full system is found
to still be too computationally expensive for real-time performance and currently
runs at 3 frames per second, however the ever improving computational power of
mobile devices and our described future work will lead to improved performance.
From this study we conclude that inertial sensors make a valuable addition into
a visual pose estimation pipeline implemented on a mobile device.AFRIKAANSE OPSOMMING: Enkel-kamera struktuur-vanaf-beweging (structure from motion, SfM) is ’n bekende
navorsingsprobleem, maar baie van die bestaande benaderings is te berekeningsintensief
vir gebruik op mobiele toestelle. In hierdie tesis ondersoek ons hoe
traagheidsensors gebruik kan word om die prestasie van SfM algoritmes op mobiele
toestelle te verbeter.
Om van die lae-koste traagheidsensors wat op meeste mobiele toestelle gevind
word gebruik te maak, ontwerp en implementeer ons ’n uitgebreide Kalman filter
(extended Kalman filter, EKF) om hul komplementêre geaardhede te ontgin, en
sodoende ’n akkurate skatting van die toestel se postuur te verkry. Ons maak van ’n
kwaternioon-gebaseerde stelselmodel gebruik om die meetstadium van die EKF te
lineariseer, en so die berekeningskompleksiteit te verminder. Hierdie afskatting van
die toestel se postuur word gebruik om die fases van kenmerkvolging en kameralokalisering
in ons SfM proses te verbeter.
Vir kenmerkvolging implementeer ons ’n hibriede volgingsalgoritme wat gebruik
maak van Harris-hoekpunte en ’n benaderde naaste-buurpunt-soektog om die
soekruimte vir moontlike ooreenstemmings te verklein. Ons verhoog die robuustheid
van hierdie benadering, deur traagheidsinligting te gebruik om vir kamerarotasies
tussen raampies te kompenseer. Verder ontwikkel ons ’n doeltreffende
bondelaanpassingsalgoritme wat slegs optimeer oor die vorige drie sleutelraampies,
en die 3D punte gemeenskaplik tussen minstens twee van hierdie raampies. Ons
implementeer ’n optimeringsgebaseerde lokaliseringsalgoritme, wat gebruik maak
van ons EKF se postuurafskatting en die gevolgde kenmerke, om die posisie en
oriëntasie van die toestel relatief tot die 3D punte in die kaart af te skat. Die optimering
word in twee stappe uitgevoer: eerstens net oor die kamera se translasie,
en tweedens oor beide die translasie en rotasie. Ons integreer die bogenoemde drie
sub-stelsels in ’n pyplyn vir postuurafskatting met behulp van traagheidsensors.
Ons evalueer ons algoritmes met die gebruik van datastelle wat met ’n iPhone
5 opgeneem is, terwyl dit in die teenwoordigheid van ’n Vicon bewegingsvasleggingstelsel
was (vir die gelyktydige opneming van korrekte postuurdata). Ons vind
dat die EKF die toestel se postuur kan afskat met ’n gemiddelde dinamiese akkuraatheid
van ±5°. Verder vind ons dat die insluiting van sensors in die visuele
postuurafskattingspyplyn kan lei tot verbeterings in terme van die robuustheid
en berekeningsdoeltreffendheid van die algoritmes, en dat dit waarskynlik nie die
akkuraatheid van so ’n stelsel negatief beïnvloed nie. Al het ons die uitvoertyd
drasties verminder (in vergelyking met tipiese bestaande tegnieke) is ons volledige
stelsel steeds te berekeningsintensief vir intydse verwerking op ’n mobiele toestel
en hardloop tans teen 3 raampies per sekonde. Die voortdurende verbetering van
mobiele toestelle se berekeningskrag en die toekomstige werk wat ons beskryf sal
egter lei tot ’n verbetering in prestasie.
Uit hierdie studie kan ons aflei dat traagheidsensors ’n waardevolle toevoeging
tot ’n visuele postuurafskattingspyplyn kan maak
Mining Hard Negative Samples for SAR-Optical Image Matching Using Generative Adversarial Networks
In this paper, we propose a generative framework to produce similar yet novel samples for a specified image. We then propose the use of these images as hard-negatives samples, within the framework of hard-negative mining, in order to improve the performance of classification networks in applications which suffer from sparse labelled training data. Our approach makes use of a variational autoencoder (VAE) which is trained in an adversarial manner in order to learn a latent distribution of the training data, as well as to be able to generate realistic, high quality image patches. We evaluate our proposed generative approach to hard-negative mining on a synthetic aperture radar (SAR) and optical image matching task. Using an existing SAR-optical matching network as the basis for our investigation, we compare the performance of the matching network trained using our approach to the baseline method, as well as to two other hard-negative mining methods. Our proposed generative architecture is able to generate realistic, very high resolution (VHR) SAR image patches which are almost indistinguishable from real imagery. Furthermore, using the patches as hard-negative samples, we are able to improve the overall accuracy, and significantly decrease the false positive rate of the SAR-optical matching task—thus validating our generative hard-negative mining approaches’ applicability to improve training in data sparse applications
A Cluster Graph Approach to Land Cover Classification Boosting
When it comes to land cover classification, the process of deriving the land classes is complex due to possible errors in algorithms, spatio-temporal heterogeneity of the Earth observation data, variation in availability and quality of reference data, or a combination of these. This article proposes a probabilistic graphical model approach, in the form of a cluster graph, to boost geospatial classifications and produce a more accurate and robust classification and uncertainty product. Cluster graphs can be characterized as a means of reasoning about geospatial data such as land cover classifications by considering the effects of spatial distribution, and inter-class dependencies in a computationally efficient manner. To assess the capabilities of our proposed cluster graph boosting approach, we apply it to the field of land cover classification. We make use of existing land cover products (GlobeLand30, CORINE Land Cover) along with data from Volunteered Geographic Information (VGI), namely OpenStreetMap (OSM), to generate a boosted land cover classification and the respective uncertainty estimates. Our approach combines qualitative and quantitative components through the application of our probabilistic graphical model and subjective expert judgments. Evaluating our approach on a test region in Garmisch-Partenkirchen, Germany, our approach was able to boost the overall land cover classification accuracy by 1.4% when compared to an independent reference land cover dataset. Our approach was shown to be robust and was able to produce a diverse, feasible and spatially consistent land cover classification in areas of incomplete and conflicting evidence. On an independent validation scene, we demonstrated that our cluster graph boosting approach was generalizable even when initialized with poor prior assumptions
A deep learning framework for matching of SAR and optical imagery
SAR and optical imagery provide highly complementary information about observed scenes. A combined use of these two modalities is thus desirable in many data fusion scenarios. However, any data fusion task requires measurements to be accurately aligned. While for both data sources images are usually provided in a georeferenced manner, the geo-localization of optical images is often inaccurate due to propagation of angular measurement errors. Many methods for the matching of homologous image regions exist for both SAR and optical imagery, however, these methods are unsuitable for SAR-optical image matching due to significant geometric and radiometric differences between the two modalities. In this paper, we present a three-step framework for sparse image matching of SAR and optical imagery, whereby each step is encoded by a deep neural network. We first predict regions in each image which are deemed most suitable for matching. A correspondence heatmap is then generated through a multi-scale, feature-space cross-correlation operator. Finally, outliers are removed by classifying the correspondence surface as a positive or negative match. Our experiments show that the proposed approach provides a substantial improvement over previous methods for SAR-optical image matching and can be used to register even large-scale scenes. This opens up the possibility of using both types of data jointly, for example for the improvement of the geo-localization of optical satellite imagery or multi-sensor stereogrammetry
Agouti C57BL/6N embryonic stem cells for mouse genetic resources.
We report the characterization of a highly germline competent C57BL/6N mouse embryonic stem cell line, JM8. To simplify breeding schemes, the dominant agouti coat color gene was restored in JM8 cells by targeted repair of the C57BL/6 nonagouti mutation. These cells provide a robust foundation for large-scale mouse knockout programs that aim to provide a public resource of targeted mutations in the C57BL/6 genetic background